Hyperfast second-order local solvers for efficient statistically preconditioned distributed optimization

نویسندگان

چکیده

Statistical preconditioning enables fast methods for distributed large-scale empirical risk minimization problems. In this approach, multiple worker nodes compute gradients in parallel, which are then used by the central node to update parameter solving an auxiliary (preconditioned) smaller-scale optimization problem. The recently proposed Statistically Preconditioned Accelerated Gradient (SPAG) method [1] has complexity bounds superior other such algorithms but requires exact solution computationally intensive problems at every iteration. paper, we propose Inexact SPAG (InSPAG) and explicitly characterize accuracy corresponding subproblem needs be solved guarantee same convergence rate as method. We build our results first developing inexact adaptive accelerated Bregman proximal gradient general under relative smoothness strong convexity assumptions, may of independent interest. Moreover, explore properties problem InSPAG algorithm assuming Lipschitz third-order derivatives strongly convexity. For class, develop a linearly convergent Hyperfast second-order estimate total with hyperfast solver. Finally, illustrate method's practical efficiency performing numerical experiments on logistic regression models. To best knowledge, these implementing high-order problems, work data where dimension is order 3 million, number samples 700 million.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Distributed Optimization with Arbitrary Local Solvers

With the growth of data and necessity for distributed optimization methods, solvers that work well on a single machine must be re-designed to leverage distributed computation. Recent work in this area has been limited by focusing heavily on developing highly specific methods for the distributed environment. These special-purpose methods are often unable to fully leverage the competitive perform...

متن کامل

Statistically Motivated Second Order Pooling

Second-order pooling, a.k.a. bilinear pooling, has proven effective for deep learning based visual recognition. However, the resulting second-order networks yield a final representation that is orders of magnitude larger than that of standard, first-order ones, making them memory-intensive and cumbersome to deploy. Here, we introduce a general, parametric compression strategy that can produce m...

متن کامل

Second order sensitivity analysis for shape optimization of continuum structures

This study focuses on the optimization of the plane structure. Sequential quadratic programming (SQP) will be utilized, which is one of the most efficient methods for solving nonlinearly constrained optimization problems. A new formulation for the second order sensitivity analysis of the two-dimensional finite element will be developed. All the second order required derivatives will be calculat...

متن کامل

Preconditioned Krylov solvers for kernel regression

A primary computational problem in kernel regression is solution of a dense linear system with the N × N kernel matrix. Because a direct solution has an O(N) cost, iterative Krylov methods are often used with fast matrix-vector products. For poorly conditioned problems, convergence of the iteration is slow and preconditioning becomes necessary. We investigate preconditioning from the viewpoint ...

متن کامل

Preconditioned Krylov solvers on GPUs

In this paper, we study the effect of enhancing GPU-accelerated Krylov solvers with preconditioners. We consider the BiCGSTAB, CGS, QMR, and IDR( s ) Krylov solvers. For a large set of test matrices, we assess the impact of Jacobi and incomplete factorization preconditioning on the solvers’ numerical stability and time-to-solution performance. We also analyze how the use of a preconditioner imp...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: EURO journal on computational optimization

سال: 2022

ISSN: ['2192-4406', '2192-4414']

DOI: https://doi.org/10.1016/j.ejco.2022.100045